Information theoretic limits of learning a sparse rule
نویسندگان
چکیده
Abstract We consider generalized linear models in regimes where the number of nonzero components signal and accessible data points are sublinear with respect to size signal. prove a variational formula for asymptotic mutual information per sample when system grows infinity. This result allows us derive an expression minimum mean-square error (MMSE) Bayesian estimator entries have discrete distribution finite support. find that, such signals suitable vanishing scalings sparsity sampling rate, MMSE is nonincreasing piecewise constant. In specific instances even displays all-or-nothing phase transition, that is, sharply jumps from its maximum value zero at critical rate. The phenomenon has previously been shown occur high-dimensional regression. Our analysis goes beyond case applies learning weights perceptron general activation function teacher-student scenario. particular, we discuss generalization set training examples.
منابع مشابه
Information-Theoretic Rule Induction
1 Background and motivation incremental learning (e.g., in decision tree design the entire tree algorithm must be re-run) while symbolic The problem of induction or "learning from exalgorithms often incorporate incremental learning as amples" can roughly be divided into two distinct cata basic mechanism. On the other hand, symbolic egories, namely the symbolic manipulation approach techniques c...
متن کاملInformation-theoretic limits of Bayesian network structure learning
In this paper, we study the informationtheoretic limits of learning the structure of Bayesian networks (BNs), on discrete as well as continuous random variables, from a finite number of samples. We show that the minimum number of samples required by any procedure to recover the correct structure grows as Ω (m) and Ω ( k logm+ k 2 /m ) for non-sparse and sparse BNs respectively, where m is the n...
متن کاملInformation-theoretic limits of Bayesian network structure learning
While there has been a lot of prior work on determining the information-theoretic limits of structure recovery in Markov random fields (MRFs), which are undirected graphical models, characterizing the information-theoretic limits of learning BNs (directed models) is important in its own right for the following reasons. First, unlike MRFs where the undirected graph corresponding to a dependence ...
متن کاملOn the Information Theoretic Limits of Learning Ising Models
We provide a general framework for computing lower-bounds on the sample complexity of recovering the underlying graphs of Ising models, given i.i.d. samples. While there have been recent results for specific graph classes, these involve fairly extensive technical arguments that are specialized to each specific graph class. In contrast, we isolate two key graph-structural ingredients that can th...
متن کاملInformation Theoretic Competitive Learning and Linguistic Rule Acquisition
In this paper, we propose a new information theoretic method for competitive learning, and demonstrate that it can discover some linguistic rules in unsupervised ways more explicitly than the traditional competitive method. The new method can directly control competitive unit activation patterns to which input-competitive connections are adjusted. This direct control of the activation patterns ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Journal of Statistical Mechanics: Theory and Experiment
سال: 2022
ISSN: ['1742-5468']
DOI: https://doi.org/10.1088/1742-5468/ac59ac